Skip to content

[Doc] Sync.#1363

Open
congw729 wants to merge 1 commit intovllm-project:mainfrom
congw729:doc/sync
Open

[Doc] Sync.#1363
congw729 wants to merge 1 commit intovllm-project:mainfrom
congw729:doc/sync

Conversation

@congw729
Copy link
Contributor

@congw729 congw729 commented Feb 13, 2026

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

Manually add features/custom_pipeline.md to .nav.yml.
Fix grammar in the PR template.
Other editions are auto-generated after mkdocs serve.

Test Plan

mkdocs serve

Test Result

Successfully launch without warning.
image


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan. Please provide the test scripts & test commands. Please state the reasons if your codes don't require additional test scripts. For test file guidelines, please check the test style doc
  • The test results. Please paste the results comparison before and after, or the e2e results.
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model. Please run mkdocs serve to sync the documentation editions to ./docs.
  • (Optional) Release notes update. If your change is user-facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

Signed-off-by: Alicia <115451386+congw729@users.noreply.github.com>
Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

Reviewed commit: a4b62c07ae

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".


The script allows overriding:
- `MODEL` (default: `Wan-AI/Wan2.2-T2V-A14B-Diffusers`)
- `PORT` (default: `8091`)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P1 Badge Fix text-to-video script default port documentation

This page says run_server.sh defaults PORT to 8091, but the actual script defaults to 8098 (examples/online_serving/text_to_video/run_server.sh), and its companion curl script also targets 8098; users who follow this doc literally will start the server on one port and send requests to another, resulting in connection failures unless they discover and override the port manually.

Useful? React with 👍 / 👎.


The script allows overriding:
- `MODEL` (default: `Wan-AI/Wan2.2-I2V-A14B-Diffusers`)
- `PORT` (default: `8091`)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P1 Badge Correct image-to-video script default port in docs

The document states run_server.sh defaults PORT to 8091, but the real script uses 8099 (examples/online_serving/image_to_video/run_server.sh) and the example curl script posts to 8099, so following this page causes users to send requests to the wrong port after launching with the provided script.

Useful? React with 👍 / 👎.

Comment on lines +27 to +28
- `BOUNDARY_RATIO` (default: `0.875`)
- `FLOW_SHIFT` (default: `12.0`)

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P2 Badge Remove unsupported image-to-video env override claims

This section claims run_server.sh supports BOUNDARY_RATIO and FLOW_SHIFT overrides, but the script does not read or pass either value (it only wires model/port/cache flags), so users setting those env vars will get silently ignored behavior and non-reproducible tuning attempts.

Useful? React with 👍 / 👎.

| `ref_audio` | string | **Yes** | Reference audio (URL or base64 data URL) |
| `ref_text` | string | No | Transcript of reference audio (for ICL mode) |
| `x_vector_only_mode` | bool | No | Use speaker embedding only (no ICL) |
| `ref_audio` | string | null | Reference audio (URL or base64 data URL) |

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P1 Badge Mark ref_audio as required for Base task

The Base-task parameter table now presents ref_audio as nullable/default-null, but the server explicitly rejects Base requests without it (vllm_omni/entrypoints/openai/serving_speech.py returns Base task requires 'ref_audio' for voice cloning), so this documentation change leads users to construct invalid requests that fail at runtime.

Useful? React with 👍 / 👎.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant